Acceleration technique for neuro symbolic integration
نویسندگان
چکیده
منابع مشابه
Acceleration Technique for Neuro Symbolic Integration
This paper presents an improved technique for accelerating the process of doing logic programming in discrete Hopfield neural network by integrating fuzzy logic and modifying activation function. Generally Hopfield networks are suitable for solving combinatorial optimization problems and pattern recognition problems. However Hopfield neural networks also face some limitations; one of the major ...
متن کاملPerspectives of Neuro–Symbolic Integration
There is an obvious tension between symbolic and subsymbolic theories, because both show complementary strengths and weaknesses in corresponding applications and underlying methodologies. The resulting gap in the foundations and the applicability of these approaches is theoretically unsatisfactory and practically undesirable. We sketch a theory that bridges this gap between symbolic and subsymb...
متن کاملApplying Fuzzy Logic in Neuro Symbolic Integration
This paper presents an improved approach for enhancing the performance of doing logic programming in Hopfield neural network. Generally Hopfield networks are suitable for solving combinatorial optimization problems. In spite of usefulness of Hopfield neural networks they have limitations; one of the most concerning drawbacks is that sometimes the solutions are local minimum instead of global mi...
متن کاملCHEBYSHEV ACCELERATION TECHNIQUE FOR SOLVING FUZZY LINEAR SYSTEM
In this paper, Chebyshev acceleration technique is used to solve the fuzzy linear system (FLS). This method is discussed in details and followed by summary of some other acceleration techniques. Moreover, we show that in some situations that the methods such as Jacobi, Gauss-Sidel, SOR and conjugate gradient is divergent, our proposed method is applicable and the acquired results are illustrate...
متن کاملNeuro-Symbolic Program Synthesis
Recent years have seen the proposal of a number of neural architectures for the problem of Program Induction. Given a set of input-output examples, these architectures are able to learn mappings that generalize to new test inputs. While achieving impressive results, these approaches have a number of important limitations: (a) they are computationally expensive and hard to train, (b) a model has...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Applied Mathematical Sciences
سال: 2015
ISSN: 1314-7552
DOI: 10.12988/ams.2015.48670